Saving Gradient and Negative Curvature Computations: Finding Local Minima More Efficiently
نویسندگان
چکیده
We propose a family of nonconvex optimization algorithms that are able to save gradient and negative curvature computations to a large extent, and are guaranteed to find an approximate local minimum with improved runtime complexity. At the core of our algorithms is the division of the entire domain of the objective function into small and large gradient regions: our algorithms only perform gradient descent based procedure in the large gradient region, and only perform negative curvature descent in the small gradient region. Our novel analysis shows that the proposed algorithms can escape the small gradient region in only one negative curvature descent step whenever they enter it, and thus they only need to perform at most N negative curvature direction computations, where N is the number of times the algorithms enter small gradient regions. For both deterministic and stochastic settings, we show that the proposed algorithms can potentially beat the state-of-the-art local minima finding algorithms. For the finite-sum setting, our algorithm can also outperform the best algorithm in a certain regime.
منابع مشابه
Third-order Smoothness Helps: Even Faster Stochastic Optimization Algorithms for Finding Local Minima
We propose stochastic optimization algorithms that can find local minima faster than existing algorithms for nonconvex optimization problems, by exploiting the third-order smoothness to escape non-degenerate saddle points more efficiently. More specifically, the proposed algorithm only needs Õ( −10/3) stochastic gradient evaluations to converge to an approximate local minimum x, which satisfies...
متن کاملNeon2: Finding Local Minima via First-Order Oracles
We propose a reduction for non-convex optimization that can (1) turn an stationary-point finding algorithm into an local-minimum finding one, and (2) replace the Hessian-vector product computations with only gradient computations. It works both in the stochastic and the deterministic settings, without hurting the algorithm’s performance. As applications, our reduction turns Natasha2 into a firs...
متن کاملShape Curvature Histogram: A Shape Feature for Celiac Disease Diagnosis
In this work we introduce a curvature based shape feature extraction technique. To construct the proposed descriptor, first an input color channel is subject to edge detection and gradient computations. Then, based on the gradient map and edge map, the local curvature of the contour is computed for each pixel as the angular difference between the maximum and minimum gradient angle within a cert...
متن کاملEstimation of Source Location Using Curvature Analysis
A quadratic surface can be fitted to potential-field data within 3×3 windows, which allow us to calculate curvature attributes from its coefficients. Phillips (2007) derived an equation depending on the most negative curvature to obtain the depth and structural index of isolated sources from peak values of special functions. They divided the special functions into two categories: Model-specific...
متن کاملStochastic Cubic Regularization for Fast Nonconvex Optimization
This paper proposes a stochastic variant of a classic algorithm—the cubic-regularized Newton method [Nesterov and Polyak, 2006]. The proposed algorithm efficiently escapes saddle points and finds approximate local minima for general smooth, nonconvex functions in only Õ( −3.5) stochastic gradient and stochastic Hessian-vector product evaluations. The latter can be computed as efficiently as sto...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- CoRR
دوره abs/1712.03950 شماره
صفحات -
تاریخ انتشار 2017